VecGAN: Image-to-Image Translation with Interpretable Latent Directions
نویسندگان
چکیده
AbstractWe propose VecGAN, an image-to-image translation framework for facial attribute editing with interpretable latent directions. Facial task faces the challenges of precise controllable strength and preservation other attributes image. For this goal, we design by space factorization each attribute, learn a linear direction that is orthogonal to others. The component change, scalar value. In our framework, can be either sampled or encoded from reference image projection. Our work inspired works fixed pretrained GANs. However, while those models cannot trained end-to-end struggle edit images precisely, VecGAN successful at preserving extensive experiments show achieves significant improvements over state-of-the-arts both local global edits.KeywordsImage translationGenerative adversarial networksLatent manipulationFace
منابع مشابه
Neural Machine Translation with Latent Semantic of Image and Text
Although attention-based Neural Machine Translation have achieved great success, attention-mechanism cannot capture the entire meaning of the source sentence because the attention mechanism generates a target word depending heavily on the relevant parts of the source sentence. The report of earlier studies has introduced a latent variable to capture the entire meaning of sentence and achieved i...
متن کاملUnsupervised Image-to-Image Translation Networks
Unsupervised image-to-image translation aims at learning a joint distribution of images in different domains by using images from the marginal distributions in individual domains. Since there exists an infinite set of joint distributions that can arrive the given marginal distributions, one could infer nothing about the joint distribution from the marginal distributions without additional assum...
متن کاملImage to Image Translation for Domain Adaptation
We propose a general framework for unsupervised domain adaptation, which allows deep neural networks trained on a source domain to be tested on a different target domain without requiring any training annotations in the target domain. This is achieved by adding extra networks and losses that help regularize the features extracted by the backbone encoder network. To this end we propose the novel...
متن کاملToward Multimodal Image-to-Image Translation
Many image-to-image translation problems are ambiguous, as a single input image may correspond to multiple possible outputs. In this work, we aim to model a distribution of possible outputs in a conditional generative modeling setting. The ambiguity of the mapping is distilled in a low-dimensional latent vector, which can be randomly sampled at test time. A generator learns to map the given inp...
متن کاملUnsupervised Image-to-Image Translation with Generative Adversarial Networks
It’s useful to automatically transform an image from its original form to some synthetic form (style, partial contents, etc.), while keeping the original structure or semantics. We define this requirement as the ”image-to-image translation” problem, and propose a general approach to achieve it, based on deep convolutional and conditional generative adversarial networks (GANs), which has gained ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture Notes in Computer Science
سال: 2022
ISSN: ['1611-3349', '0302-9743']
DOI: https://doi.org/10.1007/978-3-031-19787-1_9